Goto

Collaborating Authors

 Amherst


II-Bench: An Image Implication Understanding Benchmark for Multimodal Large Language Models

Neural Information Processing Systems

In response, numerous challenging and comprehensive benchmarks have been proposed to more accurately assess the capabilities of MLLMs. However, there is a dearth of exploration of the higher-order perceptual capabilities of MLLMs.





Self-reflecting Large Language Models: A Hegelian Dialectical Approach

Abdali, Sara, Goksen, Can, Amizadeh, Saeed, Koishida, Kazuhito

arXiv.org Artificial Intelligence

Investigating NLP through a philosophical lens has recently caught researcher's eyes as it connects computational methods with classical schools of philosophy. This paper introduces a philosophical approach inspired by the Hegelian Dialectic for LLMs' self-reflection, utilizing a self-dialectical approach to emulate internal critiques and then synthesize new ideas by resolving the contradicting points. Moreover, this paper investigates the effect of LLMs' temperature for generation by establishing a dynamic annealing approach, which promotes the creativity in the early stages and gradually refines it by focusing on the nuances, as well as a fixed temperature strategy for generation. Our proposed approach is examined to determine its ability to generate novel ideas from an initial proposition. Additionally, a Multi Agent Majority Voting (MAMV) strategy is leveraged to assess the validity and novelty of the generated ideas, which proves beneficial in the absence of domain experts. Our experiments show promise in generating new ideas and provide a stepping stone for future research.


Sharper Bounds for Chebyshev Moment Matching with Applications to Differential Privacy and Beyond

Musco, Cameron, Musco, Christopher, Rosenblatt, Lucas, Singh, Apoorv Vikram

arXiv.org Artificial Intelligence

We study the problem of approximately recovering a probability distribution given noisy measurements of its Chebyshev polynomial moments. We sharpen prior work, proving that accurate recovery in the Wasserstein distance is possible with more noise than previously known. As a main application, our result yields a simple "linear query" algorithm for constructing a differentially private synthetic data distribution with Wasserstein-1 error $\tilde{O}(1/n)$ based on a dataset of $n$ points in $[-1,1]$. This bound is optimal up to log factors and matches a recent breakthrough of Boedihardjo, Strohmer, and Vershynin [Probab. Theory. Rel., 2024], which uses a more complex "superregular random walk" method to beat an $O(1/\sqrt{n})$ accuracy barrier inherent to earlier approaches. We illustrate a second application of our new moment-based recovery bound in numerical linear algebra: by improving an approach of Braverman, Krishnan, and Musco [STOC 2022], our result yields a faster algorithm for estimating the spectral density of a symmetric matrix up to small error in the Wasserstein distance.


The Media Bias Taxonomy: A Systematic Literature Review on the Forms and Automated Detection of Media Bias

Spinde, Timo, Hinterreiter, Smi, Haak, Fabian, Ruas, Terry, Giese, Helge, Meuschke, Norman, Gipp, Bela

arXiv.org Artificial Intelligence

The way the media presents events can significantly affect public perception, which in turn can alter people's beliefs and views. Media bias describes a one-sided or polarizing perspective on a topic. This article summarizes the research on computational methods to detect media bias by systematically reviewing 3140 research papers published between 2019 and 2022. To structure our review and support a mutual understanding of bias across research domains, we introduce the Media Bias Taxonomy, which provides a coherent overview of the current state of research on media bias from different perspectives. We show that media bias detection is a highly active research field, in which transformer-based classification approaches have led to significant improvements in recent years. These improvements include higher classification accuracy and the ability to detect more fine-granular types of bias. However, we have identified a lack of interdisciplinarity in existing projects, and a need for more awareness of the various types of media bias to support methodologically thorough performance evaluations of media bias detection systems. Concluding from our analysis, we see the integration of recent machine learning advancements with reliable and diverse bias assessment strategies from other research areas as the most promising area for future research contributions in the field.


A Reinforcement Learning Approach to Optimize Available Network Bandwidth Utilization

Jamil, Hasibul, Rodrigues, Elvis, Goldverg, Jacob, Kosar, Tevfik

arXiv.org Artificial Intelligence

Efficient data transfers over high-speed, long-distance shared networks require proper utilization of available network bandwidth. Using parallel TCP streams enables an application to utilize network parallelism and can improve transfer throughput; however, finding the optimum number of parallel TCP streams is challenging due to nondeterministic background traffic sharing the same network. Additionally, the non-stationary, multi-objectiveness, and partially-observable nature of network signals in the host systems add extra complexity in finding the current network condition. In this work, we present a novel approach to finding the optimum number of parallel TCP streams using deep reinforcement learning (RL). We devise a learning-based algorithm capable of generalizing different network conditions and utilizing the available network bandwidth intelligently. Contrary to rule-based heuristics that do not generalize well in unknown network scenarios, our RL-based solution can dynamically discover and adapt the parallel TCP stream numbers to maximize the network bandwidth utilization without congesting the network and ensure fairness among contending transfers. We extensively evaluated our RL-based algorithm's performance, comparing it with several state-of-the-art online optimization algorithms. The results show that our RL-based algorithm can find near-optimal solutions 40% faster while achieving up to 15% higher throughput. We also show that, unlike a greedy algorithm, our devised RL-based algorithm can avoid network congestion and fairly share the available network resources among contending transfers.


Optimal Pose Estimation and Covariance Analysis with Simultaneous Localization and Mapping Applications

Maleki, Saeed, Raman, Adhiti, Cheng, Yang, Crassidis, John, Schmid, Matthias

arXiv.org Artificial Intelligence

This work provides a theoretical analysis for optimally solving the pose estimation problem using total least squares for vector observations from landmark features, which is central to applications involving simultaneous localization and mapping. First, the optimization process is formulated with observation vectors extracted from point-cloud features. Then, error-covariance expressions are derived. The attitude and position estimates obtained via the derived optimization process are proven to reach the bounds defined by the Cram\'er-Rao lower bound under the small-angle approximation of attitude errors. A fully populated observation noise-covariance matrix is assumed as the weight in the cost function to cover the most general case of the sensor uncertainty. This includes more generic correlations in the errors than previous cases involving an isotropic noise assumption. The proposed solution is verified using Monte Carlo simulations and an experiment with an actual LIDAR to validate the error-covariance analysis.


Blind Spots in AI Ethics and Biases in AI governance

Corrêa, Nicholas Kluge

arXiv.org Artificial Intelligence

There is an interesting link between critical theory and certain genres of literature that may be of interest to the current debate on AI ethics. While critical theory generally points out certain deficiencies in the present to criticize it, futurology and literary genres such as Cyberpunk, extrapolate our present deficits in possible dystopian futures to criticize the status quo. Given the great advance of the AI industry in recent years, an increasing number of ethical matters have been raised and debated, usually in the form of ethical guidelines and unpublished manuscripts by governments, the private sector, and academic sources. However, recent meta-analyses in the field of AI ethics have raised important questions such as: what is being omitted from published ethical guidelines? Does AI governance occur inclusively and diversely? Is this form of "ethics", based on soft rules and principles, efficient? In this study, I would like to present aspects omitted or barely mentioned in the current debate on AI ethics and defend the point that applied ethics should not be based on creating only soft versions of real legislation, but rather on criticizing the status quo for everything of value that is disregarded.